193 research outputs found
Recommended from our members
Very low bit-rate video coding focusing on moving regions using three-tier arbitrary-shaped pattern selection algorithm
Very low bit-rate video coding using patterns to represent moving regions in macroblocks exhibits good potential for improved coding efficiency. Recently an Arbitrary Shaped Pattern Selection (ASPS) algorithm and its Extended version(EASPS) were presented, that used a dynamically extracted set of patterns, of the two different sizes, based on actual video content. These algorithms, like other pattern matching algorithms failed to capture a large number of active-region macroblocks (RMB) especially when the object moving regions is relatively larger in a video sequence. As the size of the moving object may vary, superior coding performance is achievable by using dynamically extracted patterns of a larger size. This paper, proposes a three-tier Arbitrary Shaped Pattern Selection (ASPS-3) algorithm that uses three different pattern sizes for very low bit ate coding. Experimental results show that ASPS-3 exhibits better performance compared with other pattern matching algorithms, including the low-bit rate video coding standard H.263
Recommended from our members
A real time generic variable pattern selection algorithm for very low bit-rate video coding
The selection of an optimal regular-shaped pattern set for very low bit-rate video coding, focusing on moving regions has been the objective of much recent research in order to try and improve bit-rate efficiency. Selecting the optimal pattern set however, is an NP hard problem. This paper presents a generic variable pattern selection (GVPS) algorithm, which introduces a pattern selection parameter that is able to control the performance in terms of computational complexity as well as bit-rate and picture quality. While using a sub-optimal variable pattern set, GVPS obtains a coding performance comparable to near-optimal algorithms, such as the k-change neighbourhood solution, while being much less computationally intensive, so that it is able to process all types of video sequences in real-time, with minimal pre-processing overheads
Rate-Distortion Modeling for Bit Rate Constrained Point Cloud Compression
As being one of the main representation formats of 3D real world and
well-suited for virtual reality and augmented reality applications, point
clouds have gained a lot of popularity. In order to reduce the huge amount of
data, a considerable amount of research on point cloud compression has been
done. However, given a target bit rate, how to properly choose the color and
geometry quantization parameters for compressing point clouds is still an open
issue. In this paper, we propose a rate-distortion model based quantization
parameter selection scheme for bit rate constrained point cloud compression.
Firstly, to overcome the measurement uncertainty in evaluating the distortion
of the point clouds, we propose a unified model to combine the geometry
distortion and color distortion. In this model, we take into account the
correlation between geometry and color variables of point clouds and derive a
dimensionless quantity to represent the overall quality degradation. Then, we
derive the relationships of overall distortion and bit rate with the
quantization parameters. Finally, we formulate the bit rate constrained point
cloud compression as a constrained minimization problem using the derived
polynomial models and deduce the solution via an iterative numerical method.
Experimental results show that the proposed algorithm can achieve optimal
decoded point cloud quality at various target bit rates, and substantially
outperform the video-rate-distortion model based point cloud compression
scheme.Comment: Accepted to IEEE Transactions on Circuits and Systems for Video
Technolog
Human-Machine Collaborative Video Coding Through Cuboidal Partitioning
Video coding algorithms encode and decode an entire video frame while feature
coding techniques only preserve and communicate the most critical information
needed for a given application. This is because video coding targets human
perception, while feature coding aims for machine vision tasks. Recently,
attempts are being made to bridge the gap between these two domains. In this
work, we propose a video coding framework by leveraging on to the commonality
that exists between human vision and machine vision applications using cuboids.
This is because cuboids, estimated rectangular regions over a video frame, are
computationally efficient, has a compact representation and object centric.
Such properties are already shown to add value to traditional video coding
systems. Herein cuboidal feature descriptors are extracted from the current
frame and then employed for accomplishing a machine vision task in the form of
object detection. Experimental results show that a trained classifier yields
superior average precision when equipped with cuboidal features oriented
representation of the current test frame. Additionally, this representation
costs 7% less in bit rate if the captured frames are need be communicated to a
receiver
Human detection in surveillance videos and its applications - a review
Detecting human beings accurately in a visual surveillance system is crucial for diverse application areas including abnormal event detection, human gait characterization, congestion analysis, person identification, gender classification and fall detection for elderly people. The first step of the detection process is to detect an object which is in motion. Object detection could be performed using background subtraction, optical flow and spatio-temporal filtering techniques. Once detected, a moving object could be classified as a human being using shape-based, texture-based or motion-based features. A comprehensive review with comparisons on available techniques for detecting human beings in surveillance videos is presented in this paper. The characteristics of few benchmark datasets as well as the future research directions on human detection have also been discussed
Efficient high-resolution video compression scheme using background and foreground layers
Video coding using dynamic background frame achieves better compression compared to the traditional techniques by encoding background and foreground separately. This process reduces coding bits for the overall frame significantly; however, encoding background still requires many bits that can be compressed further for achieving better coding efficiency. The cuboid coding framework has been proven to be one of the most effective methods of image compression which exploits homogeneous pixel correlation within a frame and has better alignment with object boundary compared to traditional block-based coding. In a video sequence, the cuboid-based frame partitioning varies with the changes of the foreground. However, since the background remains static for a group of pictures, the cuboid coding exploits better spatial pixel homogeneity. In this work, the impact of cuboid coding on the background frame for high-resolution videos (Ultra-High-Definition (UHD) and 360-degree videos) is investigated using the multilayer framework of SHVC. After the cuboid partitioning, the method of coarse frame generation has been improved with a novel idea by keeping human-visual sensitive information. Unlike the traditional SHVC scheme, in the proposed method, cuboid coded background and the foreground are encoded in separate layers in an implicit manner. Simulation results show that the proposed video coding method achieves an average BD-Rate reduction of 26.69% and BD-PSNR gain of 1.51 dB against SHVC with significant encoding time reduction for both UHD and 360 videos. It also achieves an average of 13.88% BD-Rate reduction and 0.78 dB BD-PSNR gain compared to the existing relevant method proposed by X. Hoang Van. © 2013 IEEE
Depth sequence coding with hierarchical partitioning and spatial-domain quantization
Depth coding in 3D-HEVC deforms object shapes due to block-level edge-approximation and lacks efficient techniques to exploit the statistical redundancy, due to the frame-level clustering tendency in depth data, for higher coding gain at near-lossless quality. This paper presents a standalone mono-view depth sequence coder, which preserves edges implicitly by limiting quantization to the spatial-domain and exploits the frame-level clustering tendency efficiently with a novel binary tree-based decomposition (BTBD) technique. The BTBD can exploit the statistical redundancy in frame-level syntax, motion components, and residuals efficiently with fewer block-level prediction/coding modes and simpler context modeling for context-adaptive arithmetic coding. Compared with the depth coder in 3D-HEVC, the proposed one has achieved significantly lower bitrate at lossless to near-lossless quality range for mono-view coding and rendered superior quality synthetic views from the depth maps, compressed at the same bitrate, and the corresponding texture frames. © 1991-2012 IEEE
- …